估计回归函数的估计是兴趣的。惩罚回归(PR)是一个统计上有效的,研究的解决问题。不幸的是,在许多情况下,发现PR问题的确切解决方案是计算难以解决的。在此稿件中,我们为这些方案提出了一种基于网格的近似解决方案(MBS)。MBS将NPR的复杂功能最小化转换为有限参数,离散凸起最小化;并允许我们利用现代凸优化的工具。我们在许多明确的示例中显示MBS的应用(包括单一和多变量回归),并探讨参数数量必须随我们的样本大小增加,以便MBS保持NPR的速率 - 最优性。我们还提供了一种有效的算法,以最小化MBS目标,同时有效地利用MB中固有的稀疏性。
translated by 谷歌翻译
许多学习任务需要观察一系列图像并做出决定。在节点之间的运输盒的运输问题中,我们展示了如何将节点网络和它们之间的流量视为图像。这些图像具有可以统计汇总的有用结构信息。使用图像压缩技术,我们将图像降低到包含我们调用地理签名的可解释的地理信息的一组数字。使用地理签名,我们学习可用于建议未来网络连接的网络结构。我们开发出一种贝叶斯增强算法,利用统计总结网络信息作为前瞻和用户决策,以加强代理人的概率决策。
translated by 谷歌翻译
While the brain connectivity network can inform the understanding and diagnosis of developmental dyslexia, its cause-effect relationships have not yet enough been examined. Employing electroencephalography signals and band-limited white noise stimulus at 4.8 Hz (prosodic-syllabic frequency), we measure the phase Granger causalities among channels to identify differences between dyslexic learners and controls, thereby proposing a method to calculate directional connectivity. As causal relationships run in both directions, we explore three scenarios, namely channels' activity as sources, as sinks, and in total. Our proposed method can be used for both classification and exploratory analysis. In all scenarios, we find confirmation of the established right-lateralized Theta sampling network anomaly, in line with the temporal sampling framework's assumption of oscillatory differences in the Theta and Gamma bands. Further, we show that this anomaly primarily occurs in the causal relationships of channels acting as sinks, where it is significantly more pronounced than when only total activity is observed. In the sink scenario, our classifier obtains 0.84 and 0.88 accuracy and 0.87 and 0.93 AUC for the Theta and Gamma bands, respectively.
translated by 谷歌翻译
As demand for large corpora increases with the size of current state-of-the-art language models, using web data as the main part of the pre-training corpus for these models has become a ubiquitous practice. This, in turn, has introduced an important challenge for NLP practitioners, as they are now confronted with the task of developing highly optimized models and pipelines for pre-processing large quantities of textual data, which implies, effectively classifying and filtering multilingual, heterogeneous and noisy data, at web scale. One of the main components of this pre-processing step for the pre-training corpora of large language models, is the removal of adult and harmful content. In this paper we explore different methods for detecting adult and harmful of content in multilingual heterogeneous web data. We first show how traditional methods in harmful content detection, that seemingly perform quite well in small and specialized datasets quickly break down when confronted with heterogeneous noisy web data. We then resort to using a perplexity based approach but with a twist: Instead of using a so-called "clean" corpus to train a small language model and then use perplexity so select the documents with low perplexity, i.e., the documents that resemble this so-called "clean" corpus the most. We train solely with adult and harmful textual data, and then select the documents having a perplexity value above a given threshold. This approach will virtually cluster our documents into two distinct groups, which will greatly facilitate the choice of the threshold for the perplexity and will also allow us to obtain higher precision than with the traditional classification methods for detecting adult and harmful content.
translated by 谷歌翻译
The number of international benchmarking competitions is steadily increasing in various fields of machine learning (ML) research and practice. So far, however, little is known about the common practice as well as bottlenecks faced by the community in tackling the research questions posed. To shed light on the status quo of algorithm development in the specific field of biomedical imaging analysis, we designed an international survey that was issued to all participants of challenges conducted in conjunction with the IEEE ISBI 2021 and MICCAI 2021 conferences (80 competitions in total). The survey covered participants' expertise and working environments, their chosen strategies, as well as algorithm characteristics. A median of 72% challenge participants took part in the survey. According to our results, knowledge exchange was the primary incentive (70%) for participation, while the reception of prize money played only a minor role (16%). While a median of 80 working hours was spent on method development, a large portion of participants stated that they did not have enough time for method development (32%). 25% perceived the infrastructure to be a bottleneck. Overall, 94% of all solutions were deep learning-based. Of these, 84% were based on standard architectures. 43% of the respondents reported that the data samples (e.g., images) were too large to be processed at once. This was most commonly addressed by patch-based training (69%), downsampling (37%), and solving 3D analysis tasks as a series of 2D tasks. K-fold cross-validation on the training set was performed by only 37% of the participants and only 50% of the participants performed ensembling based on multiple identical models (61%) or heterogeneous models (39%). 48% of the respondents applied postprocessing steps.
translated by 谷歌翻译
Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these capabilities have led to widespread adoption, most LLMs are developed by resource-rich organizations and are frequently kept from the public. As a step towards democratizing this powerful technology, we present BLOOM, a 176B-parameter open-access language model designed and built thanks to a collaboration of hundreds of researchers. BLOOM is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total). We find that BLOOM achieves competitive performance on a wide variety of benchmarks, with stronger results after undergoing multitask prompted finetuning. To facilitate future research and applications using LLMs, we publicly release our models and code under the Responsible AI License.
translated by 谷歌翻译
安全可靠的自主驾驶堆栈(AD)的设计是我们时代最具挑战性的任务之一。预计这些广告将在具有完全自主权的高度动态环境中驱动,并且比人类更大的可靠性。从这个意义上讲,要高效,安全地浏览任意复杂的流量情景,广告必须具有预测周围参与者的未来轨迹的能力。当前的最新模型通常基于复发,图形和卷积网络,在车辆预测的背景下取得了明显的结果。在本文中,我们探讨了在生成模型进行运动预测中注意力的影响,考虑到物理和社会环境以计算最合理的轨迹。我们首先使用LSTM网络对过去的轨迹进行编码,该网络是计算社会背景的多头自我发言模块的输入。另一方面,我们制定了一个加权插值来计算最后一个观测框中的速度和方向,以便计算可接受的目标点,从HDMAP信息的可驱动的HDMAP信息中提取,这代表了我们的物理环境。最后,我们的发电机的输入是从多元正态分布采样的白噪声矢量,而社会和物理环境则是其条件,以预测可行的轨迹。我们使用Argoverse运动预测基准1.1验证我们的方法,从而实现竞争性的单峰结果。
translated by 谷歌翻译
近年来,深度学习(DL)算法的使用改善了基于视觉的空间应用的性能。但是,生成大量的注释数据来培训这些DL算法已被证明具有挑战性。虽然可以使用合成生成的图像,但在实际环境中测试时,经过合成数据训练的DL模型通常容易受到性能降解。在这种情况下,卢森堡大学的安全,可靠性和信任(SNT)跨学科中心开发了“ SNT Zero-G Lab”,用于在模拟现实世界太空环境的条件下培训和验证基于视觉的空间算法。 SNT Zero-G实验室开发的一个重要方面是设备选择。从实验室开发过程中学到的经验教训,本文提出了一种系统的方法,将市场调查和设备选择的实验分析结合在一起。特别是,本文专注于太空实验室中的图像采集设备:背景材料,相机和照明灯。实验分析的结果表明,在太空实验室开发项目中选择有效的设备选择需要通过实验分析来称赞的市场调查。
translated by 谷歌翻译
用于图像分类的深神经网络通常使用卷积过滤器来提取区分特征,然后再将其传递到线性分类器。大多数可解释性文献都集中在为卷积过滤器提供语义含义,以解释模型的推理过程,并确认其从输入域中使用相关信息。可以通过使用单数值分解分解其重量矩阵来研究完全连接的层,实际上研究每个矩阵中的行之间的相关性以发现地图的动力学。在这项工作中,我们为卷积层的重量张量定义了一个奇异的值分解,该分解器提供了对过滤器之间的相关性的类似理解,从而揭示了卷积图的动力学。我们使用随机矩阵理论中的最新结果来验证我们的定义。通过在图像分类网络的线性层上应用分解,我们建议一个框架,可以使用HyperGraphs应用可解释性方法来模型类别分离。我们没有寻找激活来解释网络,而是使用每个线性层具有最大相应奇异值的奇异向量来识别对网络最重要的特征。我们用示例说明了我们的方法,并介绍了本研究使用的分析工具DeepDataProfiler库。
translated by 谷歌翻译
我们提出了Theseus,这是一个有效的应用程序不合时宜的开源库,用于在Pytorch上构建的可区分非线性最小二乘(DNL)优化,为机器人技术和视觉中的端到端结构化学习提供了一个共同的框架。现有的DNLS实施是特定应用程序的,并且并不总是纳入许多对效率重要的成分。 Theseus是应用程序不可静止的,正如我们使用的几个示例应用程序所用的,这些应用程序是使用相同的基础可区分组件构建的,例如二阶优化器,标准成本功能和Lie组。为了提高效率,TheseUS纳入了对稀疏求解器,自动矢量化,批处理,GPU加速度和梯度计算的支持,并具有隐式分化和直接损耗最小化。我们在一组应用程序中进行了广泛的性能评估,显示出这些功能时显示出明显的效率提高和更好的可扩展性。项目页面:https://sites.google.com/view/theseus-ai
translated by 谷歌翻译